33 research outputs found
ROTUNDE - A Smart Meeting Cinematography Initiative: Tools, Datasets, and Benchmarks for Cognitive Interpretation and Control
We construe smart meeting cinematography with a focus on professional
situations such as meetings and seminars, possibly conducted in a distributed
manner across socio-spatially separated groups. The basic objective in smart
meeting cinematography is to interpret professional interactions involving
people, and automatically produce dynamic recordings of discussions, debates,
presentations etc in the presence of multiple communication modalities. Typical
modalities include gestures (e.g., raising one's hand for a question,
applause), voice and interruption, electronic apparatus (e.g., pressing a
button), movement (e.g., standing-up, moving around) etc. ROTUNDE, an instance
of smart meeting cinematography concept, aims to: (a) develop
functionality-driven benchmarks with respect to the interpretation and control
capabilities of human-cinematographers, real-time video editors, surveillance
personnel, and typical human performance in everyday situations; (b) Develop
general tools for the commonsense cognitive interpretation of dynamic scenes
from the viewpoint of visuo-spatial cognition centred perceptual
narrativisation. Particular emphasis is placed on declarative representations
and interfacing mechanisms that seamlessly integrate within large-scale
cognitive (interaction) systems and companion technologies consisting of
diverse AI sub-components. For instance, the envisaged tools would provide
general capabilities for high-level commonsense reasoning about space, events,
actions, change, and interaction.Comment: Appears in AAAI-2013 Workshop on: Space, Time, and Ambient
Intelligence (STAMI 2013
Grounding Dynamic Spatial Relations for Embodied (Robot) Interaction
This paper presents a computational model of the processing of dynamic
spatial relations occurring in an embodied robotic interaction setup. A
complete system is introduced that allows autonomous robots to produce and
interpret dynamic spatial phrases (in English) given an environment of moving
objects. The model unites two separate research strands: computational
cognitive semantics and on commonsense spatial representation and reasoning.
The model for the first time demonstrates an integration of these different
strands.Comment: in: Pham, D.-N. and Park, S.-B., editors, PRICAI 2014: Trends in
Artificial Intelligence, volume 8862 of Lecture Notes in Computer Science,
pages 958-971. Springe
Cognitive Interpretation of Everyday Activities - Toward Perceptual Narrative Based Visuo-Spatial Scene Interpretation
We position a narrative-centred computational model for high-level knowledge representation and reasoning in the context of a range of assistive technologies concerned with visuo-spatial perception and cognition tasks. Our proposed narrative model encompasses aspects such as space, events, actions, change, and interaction from the viewpoint of commonsense reasoning and learning in large-scale cognitive systems. The broad focus of this paper is on the domain of human-activity interpretation in smart environments, ambient intelligence etc. In the backdrop of a smart meeting cinematography domain, we position the proposed narrative model, preliminary work on perceptual narrativisation, and the immediate outlook on constructing general-purpose open-source tools for perceptual narrativisation